Goto

Collaborating Authors

 shrink deep learning model


A Foolproof Way to Shrink Deep Learning Models

#artificialintelligence

Researchers have proposed a technique for shrinking deep learning models that they say is simpler and produces more accurate results than state-of-the-art methods. Massachusetts Institute of Technology (MIT) researchers have proposed a technique for compressing deep learning models, by retraining a smaller model whose weakest connections have been "pruned," at its faster, initial rate of learning. The technique's groundwork was partly laid by the AutoML for model compression (AMC) algorithm from MIT's Song Han, which automatically removes redundant neurons and connections, and retrains the model to reinstate its initial accuracy. MIT's Jonathan Frankle and Michael Carbin determined that the model could simply be rewound to its early training rate without tinkering with any parameters. Although greater shrinkage is accompanied by reduced model accuracy, in comparing their method to AMC or earlier work by Frankle on weight-rewinding techniques, Frankle and Carbin found that it performed better regardless of the amount of compression.


A foolproof way to shrink deep learning models

#artificialintelligence

As more artificial intelligence applications move to smartphones, deep learning models are getting smaller to allow apps to run faster and save battery power. Now, MIT researchers have a new and better way to compress models. It's so simple that they unveiled it in a tweet last month: Train the model, prune its weakest connections, retrain the model at its fast, early training rate, and repeat, until the model is as tiny as you want. "That's it," says Alex Renda, a PhD student at MIT. "The standard things people do to prune their models are crazy complicated." Renda discussed the technique when the International Conference of Learning Representations (ICLR) convened remotely this month.


A foolproof way to shrink deep learning models

#artificialintelligence

As more artificial intelligence applications move to smartphones, deep learning models are getting smaller to allow apps to run faster and save battery power. Now, MIT researchers have a new and better way to compress models. It's so simple that they unveiled it in a tweet last month: Train the model, prune its weakest connections, retrain the model at its fast, early training rate, and repeat, until the model is as tiny as you want. "That's it," says Alex Renda, a PhD student at MIT. "The standard things people do to prune their models are crazy complicated." Renda discussed the technique when the International Conference of Learning Representations (ICLR) convened remotely this month.